Decentralized Learning With Lazy and Approximate Dual Gradients
نویسندگان
چکیده
This paper develops algorithms for decentralized machine learning over a network, where data are distributed, computation is localized, and communication restricted between neighbors. A line of recent research in this area focuses on improving both complexities. The methods SSDA MSDA (Scaman et al., 2017) have optimal complexity when the objective smooth strongly convex, simple to derive. However, they require solving subproblem at each step, so required accuracy solutions total computational complexities uncertain. We propose new that instead subproblem, run warm-started Katyusha small, fixed number steps. In addition, previous information sufficiently useful, local rule will decide even skip round communication, leading extra savings. show our efficient provably reducing MSDA. numerical experiments, achieve significant reduction compared with state-of-the-art.
منابع مشابه
Dual Control for Approximate Bayesian Reinforcement Learning
Control of non-episodic, finite-horizon dynamical systems with uncertain dynamics poses a tough and elementary case of the exploration-exploitation trade-off. Bayesian reinforcement learning, reasoning about the effect of actions and future observations, offers a principled solution, but is intractable. We review, then extend an old approximate approach from control theory—where the problem is ...
متن کاملLearning Efficiently with Approximate Inference via Dual Losses
Many structured prediction tasks involve complex models where inference is computationally intractable, but where it can be well approximated using a linear programming relaxation. Previous approaches for learning for structured prediction (e.g., cuttingplane, subgradient methods, perceptron) repeatedly make predictions for some of the data points. These approaches are computationally demanding...
متن کاملApproximate distance fields with non-vanishing gradients
For a given set of points S, a Euclidean distance field is defined by associating with every point p of Euclidean space Ed a value that is equal to the Euclidean distance from p to S. Such distance fields have numerous computational applications, but are expensive to compute and may not be sufficiently smooth for some applications. Instead, popular implicit modeling techniques rely on various a...
متن کاملApproximate Decentralized Bayesian Inference
This paper presents an approximate method for performing Bayesian inference in models with conditional independence over a decentralized network of learning agents. The method first employs variational inference on each individual learning agent to generate a local approximate posterior, the agents transmit their local posteriors to other agents in the network, and finally each agent combines i...
متن کاملDecentralized Consensus Algorithm with Delayed and Stochastic Gradients
Abstract. We analyze the convergence of a decentralized consensus algorithm with delayed gradient information across the network. The nodes in the network privately hold parts of the objective function and collaboratively solve for the consensus optimal solution of the total objective while they can only communicate with their immediate neighbors. In real-world networks, it is often difficult a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Signal Processing
سال: 2021
ISSN: ['1053-587X', '1941-0476']
DOI: https://doi.org/10.1109/tsp.2021.3056915